Sparse Group Feature Selection by Weighted Thresholding Homotopy Method
نویسندگان
چکیده
منابع مشابه
Iterative Hard Thresholding for Weighted Sparse Approximation
Recent work by Rauhut and Ward developed a notion of weighted sparsity and a corresponding notion of Restricted Isometry Property for the space of weighted sparse signals. Using these notions, we pose a best weighted sparse approximation problem, i.e. we seek structured sparse solutions to underdetermined systems of linear equations. Many computationally efficient greedy algorithms have been de...
متن کاملIterative Soft/Hard Thresholding Homotopy Algorithm for Sparse Recovery
In this note, we analyze an iterative soft / hard thresholding algorithm with homotopy continuation for recovering a sparse signal x† from noisy data of a noise level . Under standard regularity and sparsity conditions, we design a path along which the algorithm will find a solution x∗ which admits sharp reconstruction error ‖x−x‖`∞ = O( ) with an iteration complexity O( ln ln ρ np), where n an...
متن کاملFeature Selection by Joint Graph Sparse Coding
This paper takes manifold learning and regression simultaneously into account to perform unsupervised spectral feature selection. We first extract the bases of the data, and then represent the data sparsely using the extracted bases by proposing a novel joint graph sparse coding model, JGSC for short. We design a new algorithm TOSC to compute the resulting objective function of JGSC, and then t...
متن کاملEfficient Sparse Group Feature Selection via Nonconvex Optimization
Sparse feature selection has been demonstrated to be effective in handling high-dimensional data. While promising, most of the existing works use convex methods, which may be suboptimal in terms of the accuracy of feature selection and parameter estimation. In this paper, we expand a nonconvex paradigm to sparse group feature selection, which is motivated by applications that require identifyin...
متن کاملFeature Selection by Higher Criticism Thresholding: Optimal Phase Diagram
We consider two-class linear classification in a high-dimensional, low-sample size setting. Only a small fraction of the features are useful, the useful features are unknown to us, and each useful feature contributes weakly to the classification decision – this setting was called the rare/weak model (RW Model) in [11]. We select features by thresholding feature z-scores. The threshold is set by...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: IEEE Access
سال: 2020
ISSN: 2169-3536
DOI: 10.1109/access.2020.2968716